2024-05-09 07:19:27
Lessons from the Use of Natural Language Inference (NLI) in Requirements Engineering Tasks
Mohamad Fazelnia, Viktoria Koscinski, Spencer Herzog, Mehdi Mirakhorli
https://arxiv.org/abs/2405.05135
Lessons from the Use of Natural Language Inference (NLI) in Requirements Engineering Tasks
Mohamad Fazelnia, Viktoria Koscinski, Spencer Herzog, Mehdi Mirakhorli
https://arxiv.org/abs/2405.05135
CourseGPT-zh: an Educational Large Language Model Based on Knowledge Distillation Incorporating Prompt Optimization
Zheyan Qu, Lu Yin, Zitong Yu, Wenbo Wang, Xing zhang
https://arxiv.org/abs/2405.04781
Enhancing Holonic Architecture with Natural Language Processing for System of Systems
Muhammad Ashfaq, Ahmed R. Sadik, Tommi Mikkonen, Muhammad Waseem, Niko M akitalo
https://arxiv.org/abs/2405.05365
Exploring the True Potential: Evaluating the Black-box Optimization Capability of Large Language Models
Beichen Huang, Xingyu Wu, Yu Zhou, Jibin Wu, Liang Feng, Ran Cheng, Kay Chen Tan
https://arxiv.org/abs/2404.06290
CourseGPT-zh: an Educational Large Language Model Based on Knowledge Distillation Incorporating Prompt Optimization
Zheyan Qu, Lu Yin, Zitong Yu, Wenbo Wang, Xing zhang
https://arxiv.org/abs/2405.04781
This https://arxiv.org/abs/2403.02308 has been replaced.
link: https://scholar.google.com/scholar?q=a
Chart What I Say: Exploring Cross-Modality Prompt Alignment in AI-Assisted Chart Authoring
Nazar Ponochevnyi, Anastasia Kuzminykh
https://arxiv.org/abs/2404.05103
Understanding Language Modeling Paradigm Adaptations in Recommender Systems: Lessons Learned and Open Challenges
Lemei Zhang, Peng Liu, Yashar Deldjoo, Yong Zheng, Jon Atle Gulla
https://arxiv.org/abs/2404.03788
🔌 Utilizing BERT for Information Retrieval: Survey, Applications, Resources, and Challenges
#ai
This https://arxiv.org/abs/2403.15417 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_ees…
Privacy-Aware Semantic Cache for Large Language Models
Waris GillVirginia Tech, USA, Mohamed ElidrisiCisco, USA, Pallavi KalapatapuCisco, USA, Ali AnwarUniversity of Minnesota, Minneapolis, USA, Muhammad Ali GulzarVirginia Tech, USA
https://arxiv.org/abs/2403.02694
This https://arxiv.org/abs/2404.05694 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csCL_…
This https://arxiv.org/abs/2405.02937 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csCL_…
This https://arxiv.org/abs/2404.13236 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csDC_…
NOVA: NoC-based Vector Unit for Mapping Attention Layers on a CNN Accelerator
Mohit Upadhyay, Rohan Juneja, Weng-Fai Wong, Li-Shiuan Peh
https://arxiv.org/abs/2405.04206
This https://arxiv.org/abs/2403.18276 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csIR_…
This https://arxiv.org/abs/2401.01508 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csSE_…
🚨 Reminder to submit your best work on #Ukrainian to the 3rd Ukrainian Natural Language Processing workshop! #nlproc #callForPapers
Extended deadline: March 4!
This https://arxiv.org/abs/2404.06162 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csCL_…
This https://arxiv.org/abs/2401.01751 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csDL_…
ProLLaMA: A Protein Large Language Model for Multi-Task Protein Language Processing
Liuzhenghao Lv, Zongying Lin, Hao Li, Yuyang Liu, Jiaxi Cui, Calvin Yu-Chian Chen, Li Yuan, Yonghong Tian
https://arxiv.org/abs/2402.16445 https://arxiv.org/pdf/2402.16445
arXiv:2402.16445v1 Announce Type: new
Abstract: Large Language Models (LLMs), including GPT-x and LLaMA2, have achieved remarkable performance in multiple Natural Language Processing (NLP) tasks. Under the premise that protein sequences constitute the protein language, Protein Large Language Models (ProLLMs) trained on protein corpora excel at de novo protein sequence generation. However, as of now, unlike LLMs in NLP, no ProLLM is capable of multiple tasks in the Protein Language Processing (PLP) field. This prompts us to delineate the inherent limitations in current ProLLMs: (i) the lack of natural language capabilities, (ii) insufficient instruction understanding, and (iii) high training resource demands. To address these challenges, we introduce a training framework to transform any general LLM into a ProLLM capable of handling multiple PLP tasks. Specifically, our framework utilizes low-rank adaptation and employs a two-stage training approach, and it is distinguished by its universality, low overhead, and scalability. Through training under this framework, we propose the ProLLaMA model, the first known ProLLM to handle multiple PLP tasks simultaneously. Experiments show that ProLLaMA achieves state-of-the-art results in the unconditional protein sequence generation task. In the controllable protein sequence generation task, ProLLaMA can design novel proteins with desired functionalities. In the protein property prediction task, ProLLaMA achieves nearly 100\% accuracy across many categories. The latter two tasks are beyond the reach of other ProLLMs. Code is available at \url{https://github.com/Lyu6PosHao/ProLLaMA}.
Characterizing Multimodal Long-form Summarization: A Case Study on Financial Reports
Tianyu Cao, Natraj Raman, Danial Dervovic, Chenhao Tan
https://arxiv.org/abs/2404.06162
Exploring the Improvement of Evolutionary Computation via Large Language Models
Jinyu Cai, Jinglue Xu, Jialong Li, Takuto Ymauchi, Hitoshi Iba, Kenji Tei
https://arxiv.org/abs/2405.02876
Zero-shot LLM-guided Counterfactual Generation for Text
Amrita Bhattacharjee, Raha Moraffah, Joshua Garland, Huan Liu
https://arxiv.org/abs/2405.04793 http…
Structural Balance in Real-World Social Networks: Incorporating Direction and Transitivity in Measuring Partial Balance
Rezvaneh Rezapour, Ly Dinh, Lan Jiang, Jana Diesner
https://arxiv.org/abs/2405.02798
Zero-shot LLM-guided Counterfactual Generation for Text
Amrita Bhattacharjee, Raha Moraffah, Joshua Garland, Huan Liu
https://arxiv.org/abs/2405.04793 http…
This https://arxiv.org/abs/2310.12357 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csSE_…
ALISA: Accelerating Large Language Model Inference via Sparsity-Aware KV Caching
Youpeng Zhao, Di Wu, Jun Wang
https://arxiv.org/abs/2403.17312 https://
ClinLinker: Medical Entity Linking of Clinical Concept Mentions in Spanish
Fernando Gallego, Guillermo L\'opez-Garc\'ia, Luis Gasco-S\'anchez, Martin Krallinger, Francisco J. Veredas
https://arxiv.org/abs/2404.06367
Navigating WebAI: Training Agents to Complete Web Tasks with Large Language Models and Reinforcement Learning
Lucas-Andre\"i Thil, Mirela Popa, Gerasimos Spanakis
https://arxiv.org/abs/2405.00516
CARE-SD: Classifier-based analysis for recognizing and eliminating stigmatizing and doubt marker labels in electronic health records: model development and validation
Drew Walker, Annie Thorne, Sudeshna Das, Jennifer Love, Hannah LF Cooper, Melvin Livingston III, Abeed Sarker
https://arxiv.org/abs/2405.05204
ChangeMamba: Remote Sensing Change Detection with Spatio-Temporal State Space Model
Hongruixuan Chen, Jian Song, Chengxi Han, Junshi Xia, Naoto Yokoya
https://arxiv.org/abs/2404.03425
This https://arxiv.org/abs/2305.14461 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csDS_…
Chat2Scenario: Scenario Extraction From Dataset Through Utilization of Large Language Model
Yongqi Zhao, Wenbo Xiao, Tomislav Mihalj, Jia Hu, Arno Eichberger
https://arxiv.org/abs/2404.16147
CARE-SD: Classifier-based analysis for recognizing and eliminating stigmatizing and doubt marker labels in electronic health records: model development and validation
Drew Walker, Annie Thorne, Sudeshna Das, Jennifer Love, Hannah LF Cooper, Melvin Livingston III, Abeed Sarker
https://arxiv.org/abs/2405.05204
This https://arxiv.org/abs/2308.06013 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csIT_…
Quantitative Tools for Time Series Analysis in Natural Language Processing: A Practitioners Guide
W. Benedikt Schmal
https://arxiv.org/abs/2404.18499 https…
Application of GPT Language Models for Innovation in Activities in University Teaching
Manuel de Buenaga, Francisco Javier Bueno
https://arxiv.org/abs/2403.14694
Dendrites endow artificial neural networks with accurate, robust and parameter-efficient learning
Spyridon Chavlis, Panayiota Poirazi
https://arxiv.org/abs/2404.03708
Navigating WebAI: Training Agents to Complete Web Tasks with Large Language Models and Reinforcement Learning
Lucas-Andre\"i Thil, Mirela Popa, Gerasimos Spanakis
https://arxiv.org/abs/2405.00516
Fine-tuning Pre-trained Named Entity Recognition Models For Indian Languages
Sankalp Bahad, Pruthwik Mishra, Karunesh Arora, Rakesh Chandra Balabantaray, Dipti Misra Sharma, Parameswari Krishnamurthy
https://arxiv.org/abs/2405.04829
Fine-tuning Pre-trained Named Entity Recognition Models For Indian Languages
Sankalp Bahad, Pruthwik Mishra, Karunesh Arora, Rakesh Chandra Balabantaray, Dipti Misra Sharma, Parameswari Krishnamurthy
https://arxiv.org/abs/2405.04829
Dendrites endow artificial neural networks with accurate, robust and parameter-efficient learning
Spyridon Chavlis, Panayiota Poirazi
https://arxiv.org/abs/2404.03708
Improving Long Text Understanding with Knowledge Distilled from Summarization Model
Yan Liu, Yazheng Yang, Xiaokang Chen
https://arxiv.org/abs/2405.04955 h…
Natural Language Processing Methods for Symbolic Music Generation and Information Retrieval: a Survey
Dinh-Viet-Toan Le, Louis Bigo, Mikaela Keller, Dorien Herremans
https://arxiv.org/abs/2402.17467
Improving Long Text Understanding with Knowledge Distilled from Summarization Model
Yan Liu, Yazheng Yang, Xiaokang Chen
https://arxiv.org/abs/2405.04955 h…
Resource Allocation in Large Language Model Integrated 6G Vehicular Networks
Chang Liu, Jun Zhao
https://arxiv.org/abs/2403.19016 https://
Finding fake reviews in e-commerce platforms by using hybrid algorithms
Mathivanan Periasamy, Rohith Mahadevan, Bagiya Lakshmi S, Raja CSP Raman, Hasan Kumar S, Jasper Jessiman
https://arxiv.org/abs/2404.06339
Chat2Scenario: Scenario Extraction From Dataset Through Utilization of Large Language Model
Yongqi Zhao, Wenbo Xiao, Tomislav Mihalj, Jia Hu, Arno Eichberger
https://arxiv.org/abs/2404.16147
mALBERT: Is a Compact Multilingual BERT Model Still Worth It?
Christophe Servan (ILES, STL), Sahar Ghannay (LISN), Sophie Rosset (LISN)
https://arxiv.org/abs/2403.18338
This https://arxiv.org/abs/2404.18255 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csCL_…
Simple Techniques for Enhancing Sentence Embeddings in Generative Language Models
Bowen Zhang, Kehua Chang, Chunping Li
https://arxiv.org/abs/2404.03921 ht…
From ChatGPT, DALL-E 3 to Sora: How has Generative AI Changed Digital Humanities Research and Services?
Jiangfeng Liu, Ziyi Wang, Jing Xie, Lei Pei
https://arxiv.org/abs/2404.18518
Comparative Analysis of Retrieval Systems in the Real World
Dmytro Mozolevskyi, Waseem AlShikh
https://arxiv.org/abs/2405.02048 https://
This https://arxiv.org/abs/2401.05200 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csHC_…
This https://arxiv.org/abs/2402.05812 has been replaced.
link: https://scholar.google.com/scholar?q=a
This https://arxiv.org/abs/2402.05812 has been replaced.
link: https://scholar.google.com/scholar?q=a
Beyond Language Models: Byte Models are Digital World Simulators
Shangda Wu, Xu Tan, Zili Wang, Rui Wang, Xiaobing Li, Maosong Sun
https://arxiv.org/abs/2402.19155
VI-OOD: A Unified Representation Learning Framework for Textual Out-of-distribution Detection
Li-Ming Zhan, Bo Liu, Xiao-Ming Wu
https://arxiv.org/abs/2404.06217
This https://arxiv.org/abs/2402.17652 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csDC_…
Using Large Language Models for Natural Language Processing Tasks in Requirements Engineering: A Systematic Guideline
Andreas Vogelsang, Jannik Fischbach
https://arxiv.org/abs/2402.13823
Neural Architecture Search for Sentence Classification with BERT
Philip Kenneweg, Sarah Schr\"oder, Barbara Hammer
https://arxiv.org/abs/2403.18547 ht…
Sentiment Analysis of Citations in Scientific Articles Using ChatGPT: Identifying Potential Biases and Conflicts of Interest
Walid Hariri
https://arxiv.org/abs/2404.01800
This https://arxiv.org/abs/2401.14295 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csCL_…
Navigator: A Decentralized Scheduler for Latency-Sensitive ML Workflows
Yuting Yang, Andrea Merlina, Weijia Song, Tiancheng Yuan, Ken Birman, Roman Vitenberg
https://arxiv.org/abs/2402.17652
This https://arxiv.org/abs/2308.11131 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csIR_…
This https://arxiv.org/abs/2403.11894 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csCL_…
LLMChain: Blockchain-based Reputation System for Sharing and Evaluating Large Language Models
Mouhamed Amine Bouchiha, Quentin Telnoff, Souhail Bakkali, Ronan Champagnat, Mourad Rabah, Micka\"el Coustaty, Yacine Ghamri-Doudane
https://arxiv.org/abs/2404.13236
This https://arxiv.org/abs/2311.01020 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csSE_…
"In-Context Learning" or: How I learned to stop worrying and love "Applied Information Retrieval"
Andrew Parry, Debasis Ganguly, Manish Chandra
https://arxiv.org/abs/2405.01116
Data Augmentation with In-Context Learning and Comparative Evaluation in Math Word Problem Solving
Gulsum Yigit, Mehmet Fatih Amasyali
https://arxiv.org/abs/2404.03938
This https://arxiv.org/abs/2310.03128 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csSE_…
AXOLOTL: Fairness through Assisted Self-Debiasing of Large Language Model Outputs
Sana Ebrahimi, Kaiwen Chen, Abolfazl Asudeh, Gautam Das, Nick Koudas
https://arxiv.org/abs/2403.00198
Large Language Model Supply Chain: A Research Agenda
Shenao Wang, Yanjie Zhao, Xinyi Hou, Haoyu Wang
https://arxiv.org/abs/2404.12736 https://
This https://arxiv.org/abs/2401.05632 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csCL_…
This https://arxiv.org/abs/2403.09891 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csCL_…
Gender Bias in Large Language Models across Multiple Languages
Jinman Zhao, Yitian Ding, Chen Jia, Yining Wang, Zifan Qian
https://arxiv.org/abs/2403.00277
Analyzing the Role of Semantic Representations in the Era of Large Language Models
Zhijing Jin, Yuen Chen, Fernando Gonzalez, Jiarui Liu, Jiayi Zhang, Julian Michael, Bernhard Sch\"olkopf, Mona Diab
https://arxiv.org/abs/2405.01502
This https://arxiv.org/abs/2404.18759 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csCL_…
RAG and RAU: A Survey on Retrieval-Augmented Language Model in Natural Language Processing
Yucheng Hu, Yuxing Lu
https://arxiv.org/abs/2404.19543 https://arxiv.org/pdf/2404.19543
arXiv:2404.19543v1 Announce Type: new
Abstract: Large Language Models (LLMs) have catalyzed significant advancements in Natural Language Processing (NLP), yet they encounter challenges such as hallucination and the need for domain-specific knowledge. To mitigate these, recent methodologies have integrated information retrieved from external resources with LLMs, substantially enhancing their performance across NLP tasks. This survey paper addresses the absence of a comprehensive overview on Retrieval-Augmented Language Models (RALMs), both Retrieval-Augmented Generation (RAG) and Retrieval-Augmented Understanding (RAU), providing an in-depth examination of their paradigm, evolution, taxonomy, and applications. The paper discusses the essential components of RALMs, including Retrievers, Language Models, and Augmentations, and how their interactions lead to diverse model structures and applications. RALMs demonstrate utility in a spectrum of tasks, from translation and dialogue systems to knowledge-intensive applications. The survey includes several evaluation methods of RALMs, emphasizing the importance of robustness, accuracy, and relevance in their assessment. It also acknowledges the limitations of RALMs, particularly in retrieval quality and computational efficiency, offering directions for future research. In conclusion, this survey aims to offer a structured insight into RALMs, their potential, and the avenues for their future development in NLP. The paper is supplemented with a Github Repository containing the surveyed works and resources for further study: https://github.com/2471023025/RALM_Survey.
This https://arxiv.org/abs/2404.19048 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csCL_…
This https://arxiv.org/abs/2311.07978 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csCL_…
This https://arxiv.org/abs/2404.18255 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csCL_…
Evaluation of Geographical Distortions in Language Models: A Crucial Step Towards Equitable Representations
R\'emy Decoupes, Roberto Interdonato, Mathieu Roche, Maguelonne Teisseire, Sarah Valentin
https://arxiv.org/abs/2404.17401
2M-NER: Contrastive Learning for Multilingual and Multimodal NER with Language and Modal Fusion
Dongsheng Wang, Xiaoqin Feng, Zeming Liu, Chuan Wang
https://arxiv.org/abs/2404.17122
This https://arxiv.org/abs/2305.12829 has been replaced.
link: https://scholar.google.com/scholar?q=a
This https://arxiv.org/abs/2403.16432 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csCL_…
This https://arxiv.org/abs/2402.08015 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csCL_…
Evaluating Class Membership Relations in Knowledge Graphs using Large Language Models
Bradley P. Allen, Paul T. Groth
https://arxiv.org/abs/2404.17000 http…
Reinforcement Retrieval Leveraging Fine-grained Feedback for Fact Checking News Claims with Black-Box LLM
Xuan Zhang, Wei Gao
https://arxiv.org/abs/2404.17283
Surveying the Dead Minds: Historical-Psychological Text Analysis with Contextualized Construct Representation (CCR) for Classical Chinese
Yuqi Chen, Sixuan Li, Ying Li, Mohammad Atari
https://arxiv.org/abs/2403.00509
This https://arxiv.org/abs/2402.14379 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csCL_…
From explainable to interpretable deep learning for natural language processing in healthcare: how far from reality?
Guangming Huang, Yunfei Long, Yingya Li, Giorgos Papanastasiou
https://arxiv.org/abs/2403.11894
Here's a Free Lunch: Sanitizing Backdoored Models with Model Merge
Ansh Arora, Xuanli He, Maximilian Mozes, Srinibas Swain, Mark Dras, Qiongkai Xu
https://arxiv.org/abs/2402.19334
Tokenization Is More Than Compression
Craig W. Schmidt, Varshini Reddy, Haoran Zhang, Alec Alameddine, Omri Uzan, Yuval Pinter, Chris Tanner
https://arxiv.org/abs/2402.18376
Prompting Towards Alleviating Code-Switched Data Scarcity in Under-Resourced Languages with GPT as a Pivot
Michelle Terblanche, Kayode Olaleye, Vukosi Marivate
https://arxiv.org/abs/2404.17216
Improving Legal Judgement Prediction in Romanian with Long Text Encoders
Mihai Masala, Traian Rebedea, Horia Velicu
https://arxiv.org/abs/2402.19170 https:…
ChatGPT Alternative Solutions: Large Language Models Survey
Hanieh Alipour, Nick Pendar, Kohinoor Roy
https://arxiv.org/abs/2403.14469 https://
Exploring Safety Generalization Challenges of Large Language Models via Code
Qibing Ren, Chang Gao, Jing Shao, Junchi Yan, Xin Tan, Wai Lam, Lizhuang Ma
https://arxiv.org/abs/2403.07865
Exploring Safety Generalization Challenges of Large Language Models via Code
Qibing Ren, Chang Gao, Jing Shao, Junchi Yan, Xin Tan, Wai Lam, Lizhuang Ma
https://arxiv.org/abs/2403.07865